Goto

Collaborating Authors

 Kuwait


Supporting Automated Fact-checking across Topics: Similarity-driven Gradual Topic Learning for Claim Detection

arXiv.org Artificial Intelligence

Selecting check-worthy claims for fact-checking is considered a crucial part of expediting the fact-checking process by filtering out and ranking the check-worthy claims for being validated among the impressive amount of claims could be found online. The check-worthy claim detection task, however, becomes more challenging when the model needs to deal with new topics that differ from those seen earlier. In this study, we propose a domain-adaptation framework for check-worthy claims detection across topics for the Arabic language to adopt a new topic, mimicking a real-life scenario of the daily emergence of events worldwide. We propose the Gradual Topic Learning (GTL) model, which builds an ability to learning gradually and emphasizes the check-worthy claims for the target topic during several stages of the learning process. In addition, we introduce the Similarity-driven Gradual Topic Learning (SGTL) model that synthesizes gradual learning with a similarity-based strategy for the target topic. Our experiments demonstrate the effectiveness of our proposed model, showing an overall tendency for improving performance over the state-of-the-art baseline across 11 out of the 14 topics under study.


Rulebreakers Challenge: Revealing a Blind Spot in Large Language Models' Reasoning with Formal Logic

arXiv.org Artificial Intelligence

Formal logic has long been applied to natural language reasoning, but this approach can sometimes lead to conclusions that, while logically entailed, are factually inconsistent with the premises or are not typically inferred by humans. This study introduces the concept of "rulebreakers", which refers to instances where logical entailment diverges from factually acceptable inference. We present RULEBREAKERS, a novel dataset for evaluating Large Language Models' (LLMs) ability to distinguish between rulebreakers and non-rulebreakers. Focusing on modus tollens and disjunctive syllogism, we assess six state-of-the-art LLMs using RULEBREAKERS, measuring their performance in terms of token-level exact accuracy and model confidence. Our findings reveal that while most models perform poorly to moderately in recognizing rulebreakers, they demonstrate a latent ability to distinguish rulebreakers when assessed by their confidence levels. Further analysis suggests that the failure to recognize rulebreakers is potentially associated with the models' world knowledge and their attention distribution patterns. This research highlights the limitation of LLMs' reasoning capabilities, and contributes to the ongoing discussion on reasoning in LLMs.


Unlearning with Control: Assessing Real-world Utility for Large Language Model Unlearning

arXiv.org Artificial Intelligence

The compelling goal of eradicating undesirable data behaviors, while preserving usual model functioning, underscores the significance of machine unlearning within the domain of large language models (LLMs). Recent research has begun to approach LLM unlearning via gradient ascent (GA) -- increasing the prediction risk for those training strings targeted to be unlearned, thereby erasing their parameterized responses. Despite their simplicity and efficiency, we suggest that GA-based methods face the propensity towards excessive unlearning, resulting in various undesirable model behaviors, such as catastrophic forgetting, that diminish their practical utility. In this paper, we suggest a set of metrics that can capture multiple facets of real-world utility and propose several controlling methods that can regulate the extent of excessive unlearning. Accordingly, we suggest a general framework to better reflect the practical efficacy of various unlearning methods -- we begin by controlling the unlearning procedures/unlearned models such that no excessive unlearning occurs and follow by the evaluation for unlearning efficacy. Our experimental analysis on established benchmarks revealed that GA-based methods are far from perfect in practice, as strong unlearning is at the high cost of hindering the model utility. We conclude that there is still a long way towards practical and effective LLM unlearning, and more efforts are required in this field.


Massively Multi-Cultural Knowledge Acquisition & LM Benchmarking

arXiv.org Artificial Intelligence

Pretrained large language models have revolutionized many applications but still face challenges related to cultural bias and a lack of cultural commonsense knowledge crucial for guiding cross-culture communication and interactions. Recognizing the shortcomings of existing methods in capturing the diverse and rich cultures across the world, this paper introduces a novel approach for massively multicultural knowledge acquisition. Specifically, our method strategically navigates from densely informative Wikipedia documents on cultural topics to an extensive network of linked pages. Leveraging this valuable source of data collection, we construct the CultureAtlas dataset, which covers a wide range of sub-country level geographical regions and ethnolinguistic groups, with data cleaning and preprocessing to ensure textual assertion sentence self-containment, as well as fine-grained cultural profile information extraction. Our dataset not only facilitates the evaluation of language model performance in culturally diverse contexts but also serves as a foundational tool for the development of culturally sensitive and aware language models. Our work marks an important step towards deeper understanding and bridging the gaps of cultural disparities in AI, to promote a more inclusive and balanced representation of global cultures in the digital domain.


A Survey of Requirements for COVID-19 Mitigation Strategies. Part I: Newspaper Clips

arXiv.org Artificial Intelligence

The COVID-19 pandemic has influenced virtually all aspects of our lives. Across the world, countries have applied various mitigation strategies for the epidemic, based on social, political, and technological instruments. We postulate that one should {identify the relevant requirements} before committing to a particular mitigation strategy. One way to achieve it is through an overview of what is considered relevant by the general public, and referred to in the media. To this end, we have collected a number of news clips that mention the possible goals and requirements for a mitigation strategy. The snippets are sorted thematically into several categories, such as health-related goals, social and political impact, civil rights, ethical requirements, and so on. In a forthcoming companion paper, we will present a digest of the requirements, derived from the news clips, and a preliminary take on their formal specification.


Empirical Loss Landscape Analysis of Neural Network Activation Functions

arXiv.org Artificial Intelligence

Activation functions play a significant role in neural network design by enabling non-linearity. The choice of activation function was previously shown to influence the properties of the resulting loss landscape. Understanding the relationship between activation functions and loss landscape properties is important for neural architecture and training algorithm design. This study empirically investigates neural network loss landscapes associated with hyperbolic tangent, rectified linear unit, and exponential linear unit activation functions. Rectified linear unit is shown to yield the most convex loss landscape, and exponential linear unit is shown to yield the least flat loss landscape, and to exhibit superior generalisation performance. The presence of wide and narrow valleys in the loss landscape is established for all activation functions, and the narrow valleys are shown to correlate with saturated neurons and implicitly regularised network configurations.


Understanding the Impact of Culture in Assessing Helpfulness of Online Reviews

arXiv.org Artificial Intelligence

Online reviews have become essential for users to make informed decisions in everyday tasks ranging from planning summer vacations to purchasing groceries and making financial investments. A key problem in using online reviews is the overabundance of online that overwhelms the users. As a result, recommendation systems for providing helpfulness of reviews are being developed. This paper argues that cultural background is an important feature that impacts the nature of a review written by the user, and must be considered as a feature in assessing the helpfulness of online reviews. The paper provides an in-depth study of differences in online reviews written by users from different cultural backgrounds and how incorporating culture as a feature can lead to better review helpfulness recommendations. In particular, we analyze online reviews originating from two distinct cultural spheres, namely Arabic and Western cultures, for two different products, hotels and books. Our analysis demonstrates that the nature of reviews written by users differs based on their cultural backgrounds and that this difference varies based on the specific product being reviewed. Finally, we have developed six different review helpfulness recommendation models that demonstrate that taking culture into account leads to better recommendations.


Future of AI: New tech will create 'digital humans,' could use more energy than all working people by 2025

FOX News

CTA Thematic Programs Director Brian Comiskey and tech strategist, author David Espindola discuss how artificial intelligence will continue to change how businesses operate and the impact on consumers. Artificial intelligence will be capable of producing higher-quality "digital humans" and could use more energy than the entire global workforce by 2025, according to experts. Brian Comiskey, the Director of Thematic Programs for the Consumer Technology Association, revealed the trade association had developed an artificial intelligence working group to navigate the new technology. With his time in thematic indexing, Comiskey noted that "responsible AI" is a growing buzzword in the financial space and has become an increasing priority for leaders across various industries. By 2025, without sustainable AI practices, AI will consume more energy than the human workforce, significantly offsetting carbon-zero gains, according to Gartner's research for their 2023 strategic predictions.


Misinformation machines? Common sense the best guard against AI chatbot 'hallucinations,' experts say

FOX News

College students Tabatha Fajardo, Jay Ram and Kyra Varnavas give their take on the development of AI in the classroom on'The Story.' Artificial intelligence experts have advised consumers to use caution and trust their instincts when encountering "hallucinations" from artificial intelligence chatbots. "The number-one piece is common sense," Kayle Gishen, chief technology officer of Florida-based tech company NeonFlux, told Fox News Digital. People should verify what they see, read or find on platforms such as ChatGPT through "established sources of information," he said. AI is prone to making mistakes -- "hallucinations" in tech terminology -- just like human sources. The word "hallucinations" refers to AI outputs "that are coherent but factually incorrect or nonsensical," said Alexander Hollingsworth of Oyova, an app developer and marketing agency in Florida.


An AI-Generated News Presenter, Fedha Welcomes You in Kuwait!

#artificialintelligence

An AI-Generated news presenter has been introduced by Kuwait News, an online news organization connected to the Kuwait Times. Fedha, the host, made her debut in a brief 13-second film during which she introduced herself in Arabic. She also solicited feedback from the audience regarding their preferred source of news. The outlet's Twitter account published the video. Additionally, Fedha will reportedly use a typical Kuwaiti accent to provide news updates on the website's social media accounts. A new revolution has begun in the media industry!